116 research outputs found
Deep Manifold Traversal: Changing Labels with Convolutional Features
Many tasks in computer vision can be cast as a "label changing" problem, where the goal is to make a semantic change to the appearance of an image or some subject in an image in order to alter the class membership. Although successful task-specific methods have been developed for some label changing applications, to date no general purpose method exists. Motivated by this we propose deep manifold traversal, a method that addresses the problem in its most general form: it first approximates the manifold of natural images then morphs a test image along a traversal path away from a source class and towards a target class while staying near the manifold throughout. The resulting algorithm is surprisingly effective and versatile. It is completely data driven, requiring only an example set of images from the desired source and target domains. We demonstrate deep manifold traversal on highly diverse label changing tasks: changing an individual's appearance (age and hair color), changing the season of an outdoor image, and transforming a city skyline towards nighttime
Deep Over-sampling Framework for Classifying Imbalanced Data
Class imbalance is a challenging issue in practical classification problems
for deep learning models as well as traditional models. Traditionally
successful countermeasures such as synthetic over-sampling have had limited
success with complex, structured data handled by deep learning models. In this
paper, we propose Deep Over-sampling (DOS), a framework for extending the
synthetic over-sampling method to exploit the deep feature space acquired by a
convolutional neural network (CNN). Its key feature is an explicit, supervised
representation learning, for which the training data presents each raw input
sample with a synthetic embedding target in the deep feature space, which is
sampled from the linear subspace of in-class neighbors. We implement an
iterative process of training the CNN and updating the targets, which induces
smaller in-class variance among the embeddings, to increase the discriminative
power of the deep representation. We present an empirical study using public
benchmarks, which shows that the DOS framework not only counteracts class
imbalance better than the existing method, but also improves the performance of
the CNN in the standard, balanced settings
An auto TCD probe design and visualization
Transcranial Doppler ultrasound (TCD) is a non-invasive ultrasound method used to examine blood circulation within the brain. During TCD, ultrasound waves are transmitted through the tissues including skull. These sound waves reflect off blood cells moving within the blood vessels, allowing the radiologist to interpret their speed and direction. In this paper, an auto TCD probe is developed to control the 2D deflection angles of the probe. The techniques of Magnetic Resonance Angiography (MRA) and Magnetic Resource Imagine (MRI) have been used to build the 3D human head model and generate the structure of cerebral arteries. The K-Nearest Neighbors (KNN) algorithm as a non-parametric method has been used for signal classification and regression of corresponding arteries . Finally, a global search and local search algorithms are used to locate the ultrasound focal zone and obtain a stronger signal efficient and more accurate result
Fast Approximate Geodesics for Deep Generative Models
The length of the geodesic between two data points along a Riemannian
manifold, induced by a deep generative model, yields a principled measure of
similarity. Current approaches are limited to low-dimensional latent spaces,
due to the computational complexity of solving a non-convex optimisation
problem. We propose finding shortest paths in a finite graph of samples from
the aggregate approximate posterior, that can be solved exactly, at greatly
reduced runtime, and without a notable loss in quality. Our approach,
therefore, is hence applicable to high-dimensional problems, e.g., in the
visual domain. We validate our approach empirically on a series of experiments
using variational autoencoders applied to image data, including the Chair,
FashionMNIST, and human movement data sets.Comment: 28th International Conference on Artificial Neural Networks, 201
Temporal Model Adaptation for Person Re-Identification
Person re-identification is an open and challenging problem in computer
vision. Majority of the efforts have been spent either to design the best
feature representation or to learn the optimal matching metric. Most approaches
have neglected the problem of adapting the selected features or the learned
model over time. To address such a problem, we propose a temporal model
adaptation scheme with human in the loop. We first introduce a
similarity-dissimilarity learning method which can be trained in an incremental
fashion by means of a stochastic alternating directions methods of multipliers
optimization procedure. Then, to achieve temporal adaptation with limited human
effort, we exploit a graph-based approach to present the user only the most
informative probe-gallery matches that should be used to update the model.
Results on three datasets have shown that our approach performs on par or even
better than state-of-the-art approaches while reducing the manual pairwise
labeling effort by about 80%
The Group Loss for Deep Metric Learning
Deep metric learning has yielded impressive results in tasks such as
clustering and image retrieval by leveraging neural networks to obtain highly
discriminative feature embeddings, which can be used to group samples into
different classes. Much research has been devoted to the design of smart loss
functions or data mining strategies for training such networks. Most methods
consider only pairs or triplets of samples within a mini-batch to compute the
loss function, which is commonly based on the distance between embeddings. We
propose Group Loss, a loss function based on a differentiable label-propagation
method that enforces embedding similarity across all samples of a group while
promoting, at the same time, low-density regions amongst data points belonging
to different groups. Guided by the smoothness assumption that "similar objects
should belong to the same group", the proposed loss trains the neural network
for a classification task, enforcing a consistent labelling amongst samples
within a class. We show state-of-the-art results on clustering and image
retrieval on several datasets, and show the potential of our method when
combined with other techniques such as ensemblesComment: Accepted to European Conference on Computer Vision (ECCV) 2020,
includes non-archival supplementary materia
Visualizing dimensionality reduction of systems biology data
One of the challenges in analyzing high-dimensional expression data is the
detection of important biological signals. A common approach is to apply a
dimension reduction method, such as principal component analysis. Typically,
after application of such a method the data is projected and visualized in the
new coordinate system, using scatter plots or profile plots. These methods
provide good results if the data have certain properties which become visible
in the new coordinate system and which were hard to detect in the original
coordinate system. Often however, the application of only one method does not
suffice to capture all important signals. Therefore several methods addressing
different aspects of the data need to be applied. We have developed a framework
for linear and non-linear dimension reduction methods within our visual
analytics pipeline SpRay. This includes measures that assist the interpretation
of the factorization result. Different visualizations of these measures can be
combined with functional annotations that support the interpretation of the
results. We show an application to high-resolution time series microarray data
in the antibiotic-producing organism Streptomyces coelicolor as well as to
microarray data measuring expression of cells with normal karyotype and cells
with trisomies of human chromosomes 13 and 21
Routine Modeling with Time Series Metric Learning
version éditeur : https://rd.springer.com/chapter/10.1007/978-3-030-30484-3_47International audienceTraditionally, the automatic recognition of human activities is performed with supervised learning algorithms on limited sets of specific activities. This work proposes to recognize recurrent activity patterns, called routines, instead of precisely defined activities. The modeling of routines is defined as a metric learning problem, and an architecture, called SS2S, based on sequence-to-sequence models is proposed to learn a distance between time series. This approach only relies on inertial data and is thus non intrusive and preserves privacy. Experimental results show that a clustering algorithm provided with the learned distance is able to recover daily routines
Person Re-identification Using Clustering Ensemble Prototypes
Abstract. This paper presents an appearance-based model to deal with the person re-identification problem. Usually in a crowded scene, it is ob-served that, the appearances of most people are similar with regard to the combination of attire. In such situation it is a difficult task to distin-guish an individual from a group of alike looking individuals and yields an ambiguity in recognition for re-identification. The proper organiza-tion of the individuals based on the appearance characteristics leads to recognize the target individual by comparing with a particular group of similar looking individuals. To reconstruct a group of individual accord-ing to their appearance is a crucial task for person re-identification. In this work we focus on unsupervised based clustering ensemble approach for discovering prototypes where each prototype represents similar set of gallery image instances. The formation of each prototype depends upon the appearance characteristics of gallery instances. The estimation of k-NN classifier is employed to specify a prototype to a given probe image. The similarity measure computation is performed between the probe and a subset of gallery images, that shares the same prototype with the probe and thus reduces the number of comparisons. Re-identification perfor-mance on benchmark datasets are presented using cumulative matching characteristic (CMC) curves.
Learning Tversky Similarity
In this paper, we advocate Tversky's ratio model as an appropriate basis for
computational approaches to semantic similarity, that is, the comparison of
objects such as images in a semantically meaningful way. We consider the
problem of learning Tversky similarity measures from suitable training data
indicating whether two objects tend to be similar or dissimilar.
Experimentally, we evaluate our approach to similarity learning on two image
datasets, showing that is performs very well compared to existing methods
- …